Cross-Validation Optimization for Structured Hessian Kernel Methods
نویسنده
چکیده
We propose a highly efficient framework for kernel multi-class models with a largeand structured set of classes, and more general for penalized likelihood kernel methods.As opposed to many previous approaches which try to decompose the fitting probleminto many smaller ones, we focus on a Newton optimization of the complete model,making use of model structure and linear conjugate gradients in order to approxi-mate Newton directions. Crucially, our learning method is based entirely on matrix-vector multiplication primitives with the kernel matrices and their derivatives, allowingstraightforward specialization to new kernels, and directing the focus of code optimiza-tion to these primitives.Kernel parameters are learned automatically by maximizing the cross-validation loglikelihood, and predictive probabilities are estimated. We demonstrate our approach onlarge scale text classification tasks with hierarchical structure on throusands of classes,achieving state-of-the-art results in an order of magnitude less time than previous work.We also discuss an extension to label sequence learning with kernel conditional randomfields.
منابع مشابه
Cross-Validation Optimization for Large Scale Structured Classification Kernel Methods
We propose a highly efficient framework for penalized likelihood kernel methods applied to multiclass models with a large, structured set of classes. As opposed to many previous approaches which try to decompose the fitting problem into many smaller ones, we focus on a Newton optimization of the complete model, making use of model structure and linear conjugate gradients in order to approximate...
متن کاملCross-Validation Optimization for Large Scale Hierarchical Classification Kernel Methods
We propose a highly efficient framework for kernel multi-class models with a large and structured set of classes. Kernel parameters are learned automatically by maximizing the cross-validation log likelihood, and predictive probabilities are estimated. We demonstrate our approach on large scale text classification tasks with hierarchical class structure, achieving state-of-the-art results in an...
متن کاملSuperlinearly convergent exact penalty projected structured Hessian updating schemes for constrained nonlinear least squares: asymptotic analysis
We present a structured algorithm for solving constrained nonlinear least squares problems, and establish its local two-step Q-superlinear convergence. The approach is based on an adaptive structured scheme due to Mahdavi-Amiri and Bartels of the exact penalty method of Coleman and Conn for nonlinearly constrained optimization problems. The structured adaptation also makes use of the ideas of N...
متن کاملKernel Learning by Unconstrained Optimization
We study the problem of learning a kernel matrix from an apriori kernel and training data. An unconstrained convex optimization formulation is proposed, with an arbitrary convex smooth loss function on kernel entries and a LogDet divergence for regularization. Since the number of variables is of order O(n), standard Newton and quasi-Newton methods are too time-consuming. An operator form Hessia...
متن کاملpiCholesky: Polynomial Interpolation of Multiple Cholesky Factors for Efficient Approximate Cross-Validation
The dominant cost in solving least-square problems using Newton’s method is often that of factorizing the Hessian matrix over multiple values of the regularization parameter (λ). We propose an efficient way to interpolate the Cholesky factors of the Hessian matrix computed over a small set of λ values. This approximation enables us to optimally minimize the hold-out error while incurring only a...
متن کامل